5,313 research outputs found

    Inference by Minimizing Size, Divergence, or their Sum

    Full text link
    We speed up marginal inference by ignoring factors that do not significantly contribute to overall accuracy. In order to pick a suitable subset of factors to ignore, we propose three schemes: minimizing the number of model factors under a bound on the KL divergence between pruned and full models; minimizing the KL divergence under a bound on factor count; and minimizing the weighted sum of KL divergence and factor count. All three problems are solved using an approximation of the KL divergence than can be calculated in terms of marginals computed on a simple seed graph. Applied to synthetic image denoising and to three different types of NLP parsing models, this technique performs marginal inference up to 11 times faster than loopy BP, with graph sizes reduced up to 98%-at comparable error in marginals and parsing accuracy. We also show that minimizing the weighted sum of divergence and size is substantially faster than minimizing either of the other objectives based on the approximation to divergence presented here.Comment: Appears in Proceedings of the Twenty-Sixth Conference on Uncertainty in Artificial Intelligence (UAI2010

    Alien Registration- Mccallum, George A. (Portland, Cumberland County)

    Get PDF
    https://digitalmaine.com/alien_docs/21484/thumbnail.jp

    Epilepsy

    Get PDF
    [Handwritten

    Alien Registration- Mccallum, George A. (Portland, Cumberland County)

    Get PDF
    https://digitalmaine.com/alien_docs/21484/thumbnail.jp

    The effect of low-energy ion-implantation on the electrical transport properties of Si-SiO2 MOSFETs

    Full text link
    Using silicon MOSFETs with thin (5nm) thermally grown SiO2 gate dielectrics, we characterize the density of electrically active traps at low-temperature after 16keV phosphorus ion-implantation through the oxide. We find that, after rapid thermal annealing at 1000oC for 5 seconds, each implanted P ion contributes an additional 0.08 plus/minus 0.03 electrically active traps, whilst no increase in the number of traps is seen for comparable silicon implants. This result shows that the additional traps are ionized P donors, and not damage due to the implantation process. We also find, using the room temperature threshold voltage shift, that the electrical activation of donors at an implant density of 2x10^12 cm^-2 is ~100%.Comment: 11 pages, 10 figure

    Degrees of Change: Understanding Academics Experiences with a Shift to Flexible Technology-Enhanced Learning in Initial Teacher Education

    Get PDF
    The implementation of technology enhanced learning in higher education is often associated with changes to academic work. This article reports on a study of staff experiences with curriculum development and teaching in multiple modes of blended and online learning in a Bachelor of Education degree. The findings indicate that the changes experienced by these teacher educators were significant but not wholesale. More specifically, the findings highlight three particular areas of change that impacted on their role as teacher educators: changed pedagogical practices, particularly in staff-student communication, interaction and relationship building with students; increasing workloads associated with flexible delivery; and changed needs for staff capacity building related to issues of quality in technology enhanced learning

    On Horizontal and Vertical Separation in Hierarchical Text Classification

    Get PDF
    Hierarchy is a common and effective way of organizing data and representing their relationships at different levels of abstraction. However, hierarchical data dependencies cause difficulties in the estimation of "separable" models that can distinguish between the entities in the hierarchy. Extracting separable models of hierarchical entities requires us to take their relative position into account and to consider the different types of dependencies in the hierarchy. In this paper, we present an investigation of the effect of separability in text-based entity classification and argue that in hierarchical classification, a separation property should be established between entities not only in the same layer, but also in different layers. Our main findings are the followings. First, we analyse the importance of separability on the data representation in the task of classification and based on that, we introduce a "Strong Separation Principle" for optimizing expected effectiveness of classifiers decision based on separation property. Second, we present Hierarchical Significant Words Language Models (HSWLM) which capture all, and only, the essential features of hierarchical entities according to their relative position in the hierarchy resulting in horizontally and vertically separable models. Third, we validate our claims on real-world data and demonstrate that how HSWLM improves the accuracy of classification and how it provides transferable models over time. Although discussions in this paper focus on the classification problem, the models are applicable to any information access tasks on data that has, or can be mapped to, a hierarchical structure.Comment: Full paper (10 pages) accepted for publication in proceedings of ACM SIGIR International Conference on the Theory of Information Retrieval (ICTIR'16

    Universal schema for entity type prediction

    Full text link
    Categorizing entities by their types is useful in many applications, including knowledge base construction, relation extraction and query intent prediction. Fine-grained entity type ontologies are especially valuable, but typically difficult to design because of unavoidable quandaries about level of detail and boundary cases. Automatically classifying entities by type is challenging as well, usually involving hand-labeling data and training a supervised predictor. This paper presents a universal schema approach to fine-grained entity type prediction. The set of types is taken as the union of textual surface patterns (e.g. appositives) and pre-defined types from available databases (e.g. Freebase) - yielding not tens or hundreds of types, but more than ten thousands of entity types, such as financier, criminologist, and musical trio. We robustly learn mutual implication among this large union by learning latent vector embeddings from probabilistic matrix factorization, thus avoiding the need for hand-labeled data. Experimental results demonstrate more than 30% reduction in error versus a traditional classification approach on predicting fine-grained entities types. © 2013 ACM
    corecore